AIbase
Home
AI Tools
AI Models
MCP
AI NEWS
EN
Model Selection
Tags
Pruning Optimization

# Pruning Optimization

Sheared LLaMA 2.7B
Apache-2.0
Sheared-LLaMA-2.7B is a lightweight language model derived from Llama-2-7b through pruning and continued pretraining, consuming only a 50B token budget.
Large Language Model Transformers
S
princeton-nlp
1,131
60
Bert Base Uncased Squadv1 X1.96 F88.3 D27 Hybrid Filled Opt V1
MIT
A question-answering model fine-tuned and optimized on SQuAD v1 based on BERT-base uncased, retaining 43% of original weights through pruning techniques, achieving 1.96x faster inference speed
Question Answering System Transformers English
B
madlag
20
0
Bert Base Uncased Squadv1 X2.01 F89.2 D30 Hybrid Rewind Opt V1
MIT
A Q&A system model fine-tuned on SQuAD v1 based on the BERT-base uncased model, optimized via the nn_pruning library, achieving 2.01x faster inference speed and a 0.69 improvement in F1 score.
Question Answering System Transformers English
B
madlag
22
0
Bert Base Uncased Sst2 Acc91.1 D37 Hybrid
Apache-2.0
This model is a text classification model fine-tuned from bert-base-uncased on the SST-2 dataset, optimized through pruning with the nn_pruning library, retaining 51% of the original model's weights while achieving 91.17% accuracy.
Text Classification Transformers English
B
echarlaix
172
0
Featured Recommended AI Models
AIbase
Empowering the Future, Your AI Solution Knowledge Base
English简体中文繁體中文にほんご
© 2025AIbase